24 research outputs found
Deformable GANs for Pose-based Human Image Generation
In this paper we address the problem of generating person images conditioned
on a given pose. Specifically, given an image of a person and a target pose, we
synthesize a new image of that person in the novel pose. In order to deal with
pixel-to-pixel misalignments caused by the pose differences, we introduce
deformable skip connections in the generator of our Generative Adversarial
Network. Moreover, a nearest-neighbour loss is proposed instead of the common
L1 and L2 losses in order to match the details of the generated image with the
target image. We test our approach using photos of persons in different poses
and we compare our method with previous work in this area showing
state-of-the-art results in two benchmarks. Our method can be applied to the
wider field of deformable object generation, provided that the pose of the
articulated object can be extracted using a keypoint detector.Comment: CVPR 2018 versio
Simulasi (gosip) infotainment dalam retorika Image (keaiban) selebritis
10 pagesInternational audienceWe present a generalization of the person-image generation task, in which a human image is generated conditioned on a target pose and a set X of source appearance images. In this way, we can exploit multiple, possibly complementary images of the same person which are usually available at training and at testing time. The solution we propose is mainly based on a local attention mechanism which selects relevant information from different source image regions, avoiding the necessity to build specific generators for each specific cardinality of X. The empirical evaluation of our method shows the practical interest of addressing the person-image generation problem in a multi-source setting
How to Make an Image More Memorable? A Deep Style Transfer Approach
Recent works have shown that it is possible to automatically predict
intrinsic image properties like memorability. In this paper, we take a step
forward addressing the question: "Can we make an image more memorable?".
Methods for automatically increasing image memorability would have an impact in
many application fields like education, gaming or advertising. Our work is
inspired by the popular editing-by-applying-filters paradigm adopted in photo
editing applications, like Instagram and Prisma. In this context, the problem
of increasing image memorability maps to that of retrieving "memorabilizing"
filters or style "seeds". Still, users generally have to go through most of the
available filters before finding the desired solution, thus turning the editing
process into a resource and time consuming task. In this work, we show that it
is possible to automatically retrieve the best style seeds for a given image,
thus remarkably reducing the number of human attempts needed to find a good
match. Our approach leverages from recent advances in the field of image
synthesis and adopts a deep architecture for generating a memorable picture
from a given input image and a style seed. Importantly, to automatically select
the best style a novel learning-based solution, also relying on deep models, is
proposed. Our experimental evaluation, conducted on publicly available
benchmarks, demonstrates the effectiveness of the proposed approach for
generating memorable images through automatic style seed selectionComment: Accepted at ACM ICMR 201
Training and Tuning Generative Neural Radiance Fields for Attribute-Conditional 3D-Aware Face Generation
Generative Neural Radiance Fields (GNeRF) based 3D-aware GANs have
demonstrated remarkable capabilities in generating high-quality images while
maintaining strong 3D consistency. Notably, significant advancements have been
made in the domain of face generation. However, most existing models prioritize
view consistency over disentanglement, resulting in limited semantic/attribute
control during generation. To address this limitation, we propose a conditional
GNeRF model incorporating specific attribute labels as input to enhance the
controllability and disentanglement abilities of 3D-aware generative models.
Our approach builds upon a pre-trained 3D-aware face model, and we introduce a
Training as Init and Optimizing for Tuning (TRIOT) method to train a
conditional normalized flow module to enable the facial attribute editing, then
optimize the latent vector to improve attribute-editing precision further. Our
extensive experiments demonstrate that our model produces high-quality edits
with superior view consistency while preserving non-target regions. Code is
available at https://github.com/zhangqianhui/TT-GNeRF.Comment: 13 page